![]() Antenna downward inclination angle measurement method based on multi-scale deep semantic segmentatio
专利摘要:
A method for measuring an antenna downtilt angle based on a multi-scale deep semantic segmentation network is disclosed, including: collecting base station antenna data by using an unmanned aerial vehicle, and labeling an acquired antenna image with a labeling tool to make a data set; calling the data set for training and debugging a model; recognizing and detecting a target antenna, performing semantic segmentation on an output image, finally obtaining a target image finally segmented, and calculating a downtilt angle of the target image. The method is highly applicable, cost-effective, and safe. 公开号:EP3680609A1 申请号:EP19856443.7 申请日:2019-03-01 公开日:2020-07-15 发明作者:designation of the inventor has not yet been filed The 申请人:Wuyi University; IPC主号:G06T7-00
专利说明:
[0001] The present disclosure relates to the field of mobile communication, and in particular, to a method for measuring an antenna downtilt angle based on a multi-scale deep semantic segmentation network. BACKGROUND [0002] Nowadays, in the era of network information, the quality of mobile communication networks is extremely important. In GSM-R construction and planning, as shown in FIG. 1, an azimuth angle and a downtilt angle of an antenna affect the coverage of signals and the interference between the signals, so the antenna needs to be timely calculated and adjusted strictly to improve the quality of network signals. [0003] There are two traditional methods to measure the antenna downtilt angle: the first one is climbing to an antenna base station manually and using a measuring instrument (a compass, a slope meter, or the like) for measurement; and the second one is installing an angle sensor on the antenna to return data. The antenna is susceptible to wind, snow and other factors, resulting in a change in the downtilt angle, so it needs to be measured regularly. For the first method, as the base station is high and the number of the antennas is larger, the manual safety hazard and workload are larger, and the practicability is low. For the second method, the installation time is long, and the antenna models are different, so the installation cost of the instruments is high and the practicability is not high. Both the two methods consume a lot of manpower and material resources and are not suitable for large-scale measurement today. SUMMARY [0004] To solve the above problems, the present disclosure aims at providing a method for measuring an antenna downtilt angle based on a multi-scale deep semantic segmentation network. The method for measuring a downtilt angle of a mobile base station antenna by calling a target detection algorithm and a semantic segmentation algorithm and using an unmanned aerial vehicle as a carrier is highly applicable, cost-effective, and safe. [0005] The technical scheme adopted by the present disclosure to solve the problems is as follows:An antenna downtilt angle measuring method based on a multi-scale deep semantic segmentation network, including: collecting image data: base station antenna data is collected by using an unmanned aerial vehicle and antenna images collected are taken as a data set; predicting a target bounding box: a target antenna in the data set is positioned, and a bounding box is predicted by logistic regression; performing target recognition and semantic segmentation: target features of the target antenna in the data set are extracted, the target features are learned and processed by an activation function, a target image is output for semantic image segmentation, and pixel points of the target image and the background are classified; and calculating an antenna downtilt angle: the width and height of an antenna box are obtained according to a border of the target image to calculate the antenna downtilt angle. [0006] Further, the collecting image data includes: locating the unmanned aerial vehicle on the top of a pole of a base station antenna, and recording the longitude and latitude (L0, W0 ) of the pole in the vertical direction; causing the unmanned aerial vehicle to fly around a point of the base station antenna, setting a flight radius of the unmanned aerial vehicle, and the unmanned aerial vehicle moving around the pole along the radius on the same horizontal plane to acquire antenna images with different attitudes and angles of a mobile base station antenna as a data set. [0007] Further, the predicting a target bounding box includes:positioning a target antenna in the antenna image, predicting a bounding box by logistic regression, first dividing the entire antenna image into N*N grids, predicting the entire antenna image after the antenna image is input, scanning each grid at a time, and starting to predict the target antenna when the center of the grid where the target antenna is located is positioned, wherein 4 coordinate values predicted for each bounding box are tx, ty, tw, and th, respectively, an upper-left offset of each target cell is (cx, cy ), box heights of the bounding boxes are px, py respectively, and the network predicts their values as:b x = σ t x + c x [0008] Further, the performing target recognition and semantic segmentation includes:performing target recognition by using a network convolutional layer for feature extraction: antenna image pixel 416*416 is input, the channel number is 3, there are 32 layers of convolution kernels, each kernel has a size of 3*3, 32 layers of convolution kernels are used to learn 32 feature maps, and for color differences of the target antenna, features of the target antenna are learned by using different convolution kernels; convolutional layer up-sampling is performed during feature extraction, and a prediction formula for object classes is as follows: PrClass i | object ∗ Pr object ∗ Pr object ∗ IOU pred truth = Pr object ∗ IOU pred truth [0009] Further, the calculating an antenna downtilt angle includes:obtaining the width x and the height y of the antenna box according to the border of the target image, and calculating a downtilt angle of the base station antenna according to a geometric relation, the downtilt angle of the base station antenna being an angle θ between the base station antenna and a vertical plane: θ = arctan x y [0010] The present disclosure has the following beneficial effects: the present disclosure adopts an antenna downtilt angle measuring method based on a multi-scale deep semantic segmentation network. The method for measuring a downtilt angle of a mobile base station antenna by calling a target detection algorithm and a semantic segmentation algorithm and using an unmanned aerial vehicle as a carrier is highly applicable, cost-effective, and safe. BRIEF DESCRIPTION OF THE DRAWINGS [0011] The present disclosure is further described below with reference to the accompanying drawings and examples.Fig. 1 is a schematic diagram of a downtilt angle of a base station antenna; Fig. 2 is a flowchart of an antenna downtilt angle measuring method based on a multi-scale deep semantic segmentation network according to an embodiment of the present disclosure; Fig. 3 is a schematic diagram of border prediction for an antenna downtilt angle measuring method based on a multi-scale deep semantic segmentation network according to an embodiment of the present disclosure; Fig. 4 is a schematic diagram of a network structure for an antenna downtilt angle measuring method based on a multi-scale deep semantic segmentation network according to an embodiment of the present disclosure; Fig. 5 is a schematic diagram of a bottleneck block for an antenna downtilt angle measuring method based on a multi-scale deep semantic segmentation network according to an embodiment of the present disclosure; Fig. 6 is a schematic diagram of standard convolution for an antenna downtilt angle measuring method based on a multi-scale deep semantic segmentation network according to an embodiment of the present disclosure; Fig. 7 is a schematic diagram of high-resolution feature extraction for an antenna downtilt angle measuring method based on a multi-scale deep semantic segmentation network according to an embodiment of the present disclosure; Fig. 8 is a schematic diagram of one-dimensional low-resolution feature extraction for an antenna downtilt angle measuring method based on a multi-scale deep semantic segmentation network according to an embodiment of the present disclosure; Fig. 9 is a schematic diagram of dilated convolution for an antenna downtilt angle measuring method based on a multi-scale deep semantic segmentation network according to an embodiment of the present disclosure; and Fig. 10 is a view of a random field for an antenna downtilt angle measuring method based on a multi-scale deep semantic segmentation network according to an embodiment of the present disclosure. DETAILED DESCRIPTION [0012] Referring to Fig. 2, an antenna downtilt angle measuring method based on a multi-scale deep semantic segmentation network is provided in an embodiment of the present disclosure, including: collecting image data: base station antenna data is collected by using an unmanned aerial vehicle and antenna images collected are taken as a data set; predicting a target bounding box: a target antenna in the data set is positioned, and a bounding box is predicted by logistic regression; performing target recognition and semantic segmentation: target features of the target antenna in the data set are extracted, the target features are learned and processed by an activation function, a target image is output for semantic image segmentation, and pixel points of the target image and the background are classified; and calculating an antenna downtilt angle: the width and height of an antenna box are obtained according to a border of the target image to calculate the antenna downtilt angle. [0013] In the embodiment, the method for measuring a downtilt angle of a mobile base station antenna by calling a target detection algorithm and a semantic segmentation algorithm and using an unmanned aerial vehicle as a carrier is highly applicable, cost-effective, and safe. [0014] Further, the step of collecting image data includes:locating the unmanned aerial vehicle on the top of a pole of a base station antenna, and recording the longitude and latitude (L0, W0 ) of the pole in the vertical direction; causing the unmanned aerial vehicle to fly around a point of the base station antenna, setting a flight radius of the unmanned aerial vehicle, and the unmanned aerial vehicle moving around the pole along the radius on the same horizontal plane to acquire antenna images with different attitudes and angles of a mobile base station antenna as a data set. [0015] Further, the step of predicting a target bounding box includes:positioning a target antenna in the antenna image, predicting a bounding box by logistic regression, first dividing the entire antenna image into N*N grids, predicting the entire antenna image after the antenna image is input, scanning each grid at a time, and starting to predict the target antenna when the center of the grid where the target antenna is located is positioned, wherein 4 coordinate values predicted for each bounding box are tx, ty, tw, and th, respectively, an upper-left offset of each target cell is (cx, cy ), box heights of the bounding boxes are px, py respectively, box prediction is as shown in Fig. 3, and the network predicts their values as:b x = σ t x + c x [0016] In the accuracy of a target, multi-scale prediction is used. There is no need to fix the size of an input image, so different step sizes can be used to detect feature maps of different sizes. Three different detection layers are used to detect the antenna image for the target antenna, and different detection layers are realized by controlling the step size. The first detection layer is down-sampled with a step size of 32 to reduce the feature dimension. In order to connect with the previous identical feature graph, the layer is up-sampled, and a high resolution can be obtained at this point. The second detection layer with a step size of 16 is used, and the remaining feature processing is consistent with that of the first layer. The step size is set to 8 in the third layer, feature prediction is performed thereon, and finally, the detection accuracy of the target antenna is greater. [0017] Further, the step of performing target recognition and semantic segmentation includes:performing target recognition by using a network convolutional layer for feature extraction: antenna image pixel 416*416 is input, the channel number is 3, there are 32 layers of convolution kernels, each kernel has a size of 3*3, 32 layers of convolution kernels are used to learn 32 feature maps, and for color differences of the target antenna, features of the target antenna are learned by using different convolution kernels; convolutional layer up-sampling is performed during feature extraction, and a prediction formula for object classes is as follows: PrClass i | object ∗ Pr object ∗ Pr object ∗ IOU pred truth = Pr object ∗ IOU pred truth [0018] a predicted target output range is made between 0 and 1, the antenna image is processed by the activation function after feature extraction, and when the output value is greater than 0.5, the object is determined as an antenna;in a network layer structure, there are 53 convolutional layers and 22 residual layers among layers 0-74; layers 75-105 are feature interaction layers of a neural convolutional network, which can be divided into three scales; local feature interaction is realized by means of convolution kernels, and its network structure is as shown in Fig. 4. [0019] In the production of the data set, only the antenna is detected, so the class is 1. Therefore, in the training, the output of the last convolutional layer is 3*(1+4+1)=18. Semantic segmentation [0020] Semantic image segmentation is performed on the antenna image by using a deep convolutional network, and the pixel points of the target image and the background are classified. [0021] After the target image is input, it first goes through feature extraction by a dilated convolutional network. Since the measured boundary precision is not high enough, the pixel of the target image cannot be well separated from the background pixel, and the pixel classification of the image boundary can be improved by combining a fully connected conditional random field, so that the segmentation effect can be better. [0022] It is first feature-extracted by using a dilated convolutional network. The feature extraction of the network convolutional layer can be divided into two cases: a low-resolution input image is feature-extracted by a standard convolutional layer, as shown in Fig. 6. Dense features of a high-resolution input image are extracted by a detour convolution at a rate of 2, as shown in Fig. 7, and its step size is set to 2 to thus reduce the feature dimension. In the convolutional network layer, the convolution kernel is set to 3, the stride length is 1, and the step size is 1. Fig. 8 is a schematic diagram of one-dimensional low-resolution feature map extraction. Fig. 9 is a schematic diagram of dilated convolution. [0023] In a network structure of a serial module and a spatial pyramid pooling layer module, the convolution with holes can effectively increase a receptive field of a filter and integrate multi-scale information. After a feature image is input, dilated convolution is calculated: y i = ∑ k x i + r ∗ k ∗ w k [0024] For a two-dimensional signal, an output corresponding to each position i is y, w is a filter, and the detour rate r is a step size for sampling the input signal. The receptive field of the filter can be improved, and the convolution with holes can enlarge the convolution kernel. A residual module of multi-scale feature learning is used in the feature network extraction, while the bottleneck block is used in the present disclosure. In the bottleneck block, each convolution is processed by normalization and processed by an activation function. Thus, contextual information of the context is enriched, and the bottleneck block is as shown in Fig. 5. [0025] After the input image is processed by the convolutional network for output, pixel points of the output target image are classified by a fully connected conditional random field, and the classification is mainly performed for the target image and the background boundary. [0026] A view of a random field is as shown in Fig. 10. Each circle represents a pixel point, xi (white circle) is a labeled pixel point (node), two connected pixel points are edges of the pixel, yi (black circle) is a reference value of xi , and the classification of the labeled pixel points is determined by the reference value yi . According to the Gibbs distribution function, P Y = y | I = 1 Z Iexp − E y | I [0027] An image function output through a dilated convolutional network is a unary potential function: ∑ψ u (yi ). A binary potential function is iΨ Py i y j = u y i y j ∑ m = 1 Mw mk G m f i f j [0028] The function the relationship between pixels and will assign the same symbols to the same prime points. The unary potential function extracts feature vectors of a node in different feature maps, and the binary function connects the nodes extracted by the unitary potential function to learn its edges. All the nodes are connected to form a conditional random field of a fully connected layer, and an image finally output by the function is more accurate. [0029] Further, the step of calculating an antenna downtilt angle includes:obtaining the width x and the height y of the antenna box according to the border of the target image, and calculating a downtilt angle of the base station antenna according to a geometric relation, the downtilt angle of the base station antenna being an angle θ between the base station antenna and a vertical plane: θ = arctan x y [0030] The above are merely preferred embodiments of the present disclosure. The present disclosure is not limited to the above implementations. As long as the implementations can achieve the technical effect of the present disclosure with the same means, they are all encompassed in the protection scope of the present disclosure.
权利要求:
Claims (5) [0001] A method for measuring an antenna downtilt angle based on a multi-scale deep semantic segmentation network, comprising: collecting image data: wherein base station antenna data is collected by using an unmanned aerial vehicle, and antenna images collected are taken as a data set; predicting a target bounding box: wherein a target antenna in the data set is positioned, and a bounding box is predicted by logistic regression; performing target recognition and semantic segmentation: wherein target features of the target antenna in the data set are extracted, the target features are learned and processed by an activation function, a target image is output for semantic image segmentation, and pixel points of the target image and the background are classified; and calculating an antenna downtilt angle: wherein the width and height of an antenna box are obtained according to a border of the target image to calculate the antenna downtilt angle. [0002] The method for measuring an antenna downtilt angle based on a multi-scale deep semantic segmentation network according to claim 1, wherein the step of collecting image data comprises: locating the unmanned aerial vehicle on the top of a pole of a base station antenna, and recording the longitude and latitude (L0, W0 ) of the pole in the vertical direction; causing the unmanned aerial vehicle to fly around a point of the base station antenna, setting a flight radius of the unmanned aerial vehicle, and the unmanned aerial vehicle moving around the pole along the radius on the same horizontal plane to acquire antenna images with different attitudes and angles of a mobile base station antenna as a data set. [0003] The method for measuring an antenna downtilt angle based on a multi-scale deep semantic segmentation network according to claim 2, wherein the step of predicting a target bounding box comprises: positioning a target antenna in the antenna image, predicting a bounding box by logistic regression, first dividing the entire antenna image into N*N grids, predicting the entire antenna image after the antenna image is input, scanning each grid at a time, and starting to predict the target antenna when the center of the grid where the target antenna is located is positioned, wherein four coordinate values predicted for each bounding box are tx, ty, tw, and th, respectively, an upper-left offset of each target cell is (cx, cy ), box heights of the bounding boxes are px, py respectively, and the network predicts their values as:b x = σ t x + c x [0004] The method for measuring an antenna downtilt angle based on a multi-scale deep semantic segmentation network according to claim 3, wherein the step of performing target recognition and semantic segmentation comprises: performing target recognition by using a network convolutional layer for feature extraction: antenna image pixel 416*416 is input, the channel number is 3, there are 32 layers of convolution kernels, each kernel has a size of 3*3, 32 layers of convolution kernels are used to learn 32 feature maps, and for color differences of the target antenna, features of the target antenna are learned by using different convolution kernels; convolutional layer up-sampling is performed during feature extraction, and a prediction formula for object classes is as follows: PrClass i | object ∗ Pr object ∗ Pr object ∗ IOU pred truth = Pr object ∗ IOU pred truth [0005] The method for measuring an antenna downtilt angle based on a multi-scale deep semantic segmentation network according to claim 4, wherein the step of calculating an antenna downtilt angle comprises:obtaining the width x and the height y of the antenna box according to the border of the target image, and calculating a downtilt angle of the base station antenna according to a geometric relation, the downtilt angle of the base station antenna being an angle θ between the base station antenna and a vertical plane: θ = arctan x y
类似技术:
公开号 | 公开日 | 专利标题 US10215571B2|2019-02-26|System and method for precision localization and mapping Liu et al.2015|Fast multiclass vehicle detection on aerial images Benjdira et al.2019|Car detection using unmanned aerial vehicles: Comparison between faster r-cnn and yolov3 Ke et al.2016|Real-time bidirectional traffic flow parameter estimation from aerial videos CN106023605B|2018-10-12|A kind of method for controlling traffic signal lights based on depth convolutional neural networks Marquez et al.2013|Intra-hour DNI forecasting based on cloud tracking image analysis CN105389567B|2019-01-25|Group abnormality detection method based on dense optical flow histogram Leitloff et al.2014|An operational system for estimating road traffic information from aerial images US10037604B2|2018-07-31|Multi-cue object detection and analysis CN105160309B|2018-12-07|Three lanes detection method based on morphological image segmentation and region growing RU2669656C2|2018-10-12|Condition detection with use of image processing CN102622607B|2013-09-25|Remote sensing image classification method based on multi-feature fusion Solberg et al.1999|Automatic detection of oil spills in ERS SAR images Fathi et al.2010|Detecting road intersections from GPS traces Wang et al.2017|SigVox–A 3D feature matching algorithm for automatic street object recognition in mobile laser scanning point clouds Liu et al.2017|A high resolution optical satellite image dataset for ship recognition and some new baselines CN105354568A|2016-02-24|Convolutional neural network based vehicle logo identification method Chen et al.2015|Vehicle detection in high-resolution aerial images via sparse representation and superpixels Zheng et al.2013|A novel vehicle detection method with high resolution highway aerial image US20120148094A1|2012-06-14|Image based detecting system and method for traffic parameters and computer program product thereof CN104166841A|2014-11-26|Rapid detection identification method for specified pedestrian or vehicle in video monitoring network Le Vu et al.2018|Angular Momentum Eddy Detection and tracking Algorithm | and its application to coastal eddy formation CN108052940A|2018-05-18|SAR remote sensing images waterborne target detection methods based on deep learning US20150235092A1|2015-08-20|Parts based object tracking method and apparatus Varadharajan et al.2014|Vision for road inspection
同族专利:
公开号 | 公开日 EP3680609A4|2021-02-24| WO2020093630A1|2020-05-14| CN109685762A|2019-04-26|
引用文献:
公开号 | 申请日 | 公开日 | 申请人 | 专利标题
法律状态:
2020-03-11| STAA| Information on the status of an ep patent application or granted ep patent|Free format text: STATUS: UNKNOWN | 2020-05-16| STAA| Information on the status of an ep patent application or granted ep patent|Free format text: STATUS: THE INTERNATIONAL PUBLICATION HAS BEEN MADE | 2020-06-12| PUAI| Public reference made under article 153(3) epc to a published international application that has entered the european phase|Free format text: ORIGINAL CODE: 0009012 | 2020-06-12| STAA| Information on the status of an ep patent application or granted ep patent|Free format text: STATUS: REQUEST FOR EXAMINATION WAS MADE | 2020-07-15| AK| Designated contracting states|Kind code of ref document: A1 Designated state(s): AL AT BE BG CH CY CZ DE DK EE ES FI FR GB GR HR HU IE IS IT LI LT LU LV MC MK MT NL NO PL PT RO RS SE SI SK SM TR | 2020-07-15| AX| Request for extension of the european patent|Extension state: BA ME | 2020-07-15| 17P| Request for examination filed|Effective date: 20200311 | 2021-01-27| RIN1| Information on inventor provided before grant (corrected)|Inventor name: ZENG, JUNYING Inventor name: DENG, WENBO Inventor name: ZHAI, YIKUI Inventor name: GAN, JUNYING Inventor name: XU, YING Inventor name: WU, YUETING Inventor name: KE, QIRUI Inventor name: ZHENG, YU Inventor name: ZHOU, JIHUA | 2021-02-24| A4| Supplementary search report drawn up and despatched|Effective date: 20210120 | 2021-02-24| RIC1| Information provided on ipc code assigned before grant|Ipc: G06K9/32 20060101ALI20210115BHEP Ipc: G01C9/00 20060101ALI20210115BHEP Ipc: G06K9/62 20060101ALI20210115BHEP Ipc: G06T7/73 20170101ALI20210115BHEP Ipc: H01Q1/12 20060101ALI20210115BHEP Ipc: G01C1/04 20060101AFI20210115BHEP Ipc: G06K9/00 20060101ALI20210115BHEP | 2022-02-16| DAV| Request for validation of the european patent (deleted)| 2022-02-16| DAX| Request for extension of the european patent (deleted)|
优先权:
[返回顶部]
申请号 | 申请日 | 专利标题 相关专利
Sulfonates, polymers, resist compositions and patterning process
Washing machine
Washing machine
Device for fixture finishing and tension adjusting of membrane
Structure for Equipping Band in a Plane Cathode Ray Tube
Process for preparation of 7 alpha-carboxyl 9, 11-epoxy steroids and intermediates useful therein an
国家/地区
|